minimal weight
On Neural Networks with Minimal Weights
Linear threshold elements are the basic building blocks of artificial neural networks. A linear threshold element computes a function that is a sign of a weighted sum of the input variables. The weights are arbitrary integers; actually, they can be very big integers-(cid:173) exponential in the number of the input variables. However, in practice, it is difficult to implement big weights. In the present literature a distinction is made between the two extreme cases: linear threshold functions with polynomial-size weights as opposed to those with exponential-size weights.
On Neural Networks with Minimal Weights
Bohossian, Vasken, Bruck, Jehoshua
Linear threshold elements are the basic building blocks of artificial neural networks. A linear threshold element computes a function that is a sign of a weighted sum of the input variables. The weights are arbitrary integers; actually, they can be very big integers-exponential in the number of the input variables. However, in practice, it is difficult to implement big weights. In the present literature a distinction is made between the two extreme cases: linear threshold functions with polynomial-size weights as opposed to those with exponential-size weights.
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > United States > California > Santa Clara County > San Jose (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > Switzerland > Vaud > Lausanne (0.04)
On Neural Networks with Minimal Weights
Bohossian, Vasken, Bruck, Jehoshua
Linear threshold elements are the basic building blocks of artificial neural networks. A linear threshold element computes a function that is a sign of a weighted sum of the input variables. The weights are arbitrary integers; actually, they can be very big integers-exponential in the number of the input variables. However, in practice, it is difficult to implement big weights. In the present literature a distinction is made between the two extreme cases: linear threshold functions with polynomial-size weights as opposed to those with exponential-size weights.
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > United States > California > Santa Clara County > San Jose (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > Switzerland > Vaud > Lausanne (0.04)
On Neural Networks with Minimal Weights
Bohossian, Vasken, Bruck, Jehoshua
A linear threshold element computes a function that is a sign of a weighted sum of the input variables. The weights are arbitrary integers; actually, they can be very big integers-exponential in the number of the input variables. However, in practice, it is difficult to implement big weights. In the present literature a distinction is made between the two extreme cases: linear threshold functions with polynomial-size weights as opposed to those with exponential-size weights. The main contribution of this paper is to fill up the gap by further refining that separation.
- North America > United States > California > Santa Clara County > Stanford (0.04)
- North America > United States > California > Santa Clara County > San Jose (0.04)
- North America > United States > California > Los Angeles County > Pasadena (0.04)
- Europe > Switzerland > Vaud > Lausanne (0.04)